No More Pesky Learning Rates: Supplementary Material

نویسندگان

  • Tom Schaul
  • Sixin Zhang
  • Yann LeCun
چکیده

If we do gradient descent with η * (t), then almost surely, the algorithm converges (for the quadratic model). To prove that, we follow classical techniques based on Lyapunov stability theory (Bucy, 1965). Notice that the expected loss follows E J θ (t+1) | θ (t) = 1 2 h · E (1 − η * h)(θ (t) − θ *) + η * hσξ 2 + σ 2 = 1 2 h (1 − η * h) 2 (θ (t) − θ *) 2 + (η *) 2 h 2 σ 2 + σ 2 = 1 2 h σ 2 (θ (t) − θ *) 2 + σ 2 (θ (t) − θ *) 2 + σ 2 ≤ J θ (t) Thus J(θ (t)) is a positive super-martingale, indicating that almost surely J(θ (t)) → J ∞. We are to prove that almost surely J ∞ = J(θ *) = 1 2 hσ 2. Observe that J(θ (t)) − E[J(θ (t+1)) | θ (t) ] = 1 2 hη * (t) , E[J(θ (t))] − E[J(θ (t+1)) | θ (t) ] = 1 2 hE[η * (t)] Since E[J(θ (t))] is bounded below by 0, the telescoping sum gives us E[η * (t)] → 0, which in turn implies that in probability η * (t) → 0. We can rewrite this as η * (t) = J(θ t) − 1 2 hσ 2 J(θ t) → 0 By uniqueness of the limit, almost surely,

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

No more pesky learning rates

The performance of stochastic gradient descent (SGD) depends critically on how learning rates are tuned and decreased over time. We propose a method to automatically adjust multiple learning rates so as to minimize the expected error at any one time. The method relies on local gradient variations across samples. In our approach, learning rates can increase as well as decrease, making it suitabl...

متن کامل

Inverse Reinforcement Learning with Simultaneous Estimation of Rewards and Dynamics - Supplementary Material

This document contains supplementary material to the paper Inverse Reinforcement Learning with Simultaneous Estimation of Rewards and Dynamics with more detailed derivations, additional proofs to lemmata and theorems as well as larger illustrations and plots of the evaluation task. 1 Partial Derivative of the Policy

متن کامل

Supplementary Materials for Learning to Recognize Discontiguous Entities

This is the supplementary material for “Learning to Recognize Discontiguous Entities” [Muis and Lu, 2016]. This material gives more details in the experiment setup, the ambiguity of each model, and compare the models from theoretical point of view.

متن کامل

Part-based Visual Tracking with Online Latent Structural Learning: Supplementary Material

In this supplementary material, we provide more related work, more details of structured output tracking and online latent structural learning for visual tracking, and more experimental results (both qualitative and quantitative). The experimental results includes demonstration videos, more quantitative results of our tracking algorithm with and without parts in Tab. 2 and Fig. 1, more quantita...

متن کامل

Identification of the Quantitative Dimensions of Supplementary Private Tutoring among High School Students

Identification of the Quantitative Dimensions of Supplementary Private Tutoring among High School Students   Y. Shiraavand* A. Ma’dandaar Aaraani, Ph.D.** I. Nikpey, Ph.D.***   To identify the quantitative dimensions of supplementary private tutoring, a cluster sample of 800 high school students was given a questionnaire on the subject. The collected data reveal that moti...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2013